141 research outputs found
Robust Localization from Incomplete Local Information
We consider the problem of localizing wireless devices in an ad-hoc network
embedded in a d-dimensional Euclidean space. Obtaining a good estimation of
where wireless devices are located is crucial in wireless network applications
including environment monitoring, geographic routing and topology control. When
the positions of the devices are unknown and only local distance information is
given, we need to infer the positions from these local distance measurements.
This problem is particularly challenging when we only have access to
measurements that have limited accuracy and are incomplete. We consider the
extreme case of this limitation on the available information, namely only the
connectivity information is available, i.e., we only know whether a pair of
nodes is within a fixed detection range of each other or not, and no
information is known about how far apart they are. Further, to account for
detection failures, we assume that even if a pair of devices is within the
detection range, it fails to detect the presence of one another with some
probability and this probability of failure depends on how far apart those
devices are. Given this limited information, we investigate the performance of
a centralized positioning algorithm MDS-MAP introduced by Shang et al., and a
distributed positioning algorithm, introduced by Savarese et al., called
HOP-TERRAIN. In particular, for a network consisting of n devices positioned
randomly, we provide a bound on the resulting error for both algorithms. We
show that the error is bounded, decreasing at a rate that is proportional to
R/Rc, where Rc is the critical detection range when the resulting random
network starts to be connected, and R is the detection range of each device.Comment: 40 pages, 13 figure
Eliminating Latent Discrimination: Train Then Mask
How can we control for latent discrimination in predictive models? How can we
provably remove it? Such questions are at the heart of algorithmic fairness and
its impacts on society. In this paper, we define a new operational fairness
criteria, inspired by the well-understood notion of omitted variable-bias in
statistics and econometrics. Our notion of fairness effectively controls for
sensitive features and provides diagnostics for deviations from fair decision
making. We then establish analytical and algorithmic results about the
existence of a fair classifier in the context of supervised learning. Our
results readily imply a simple, but rather counter-intuitive, strategy for
eliminating latent discrimination. In order to prevent other features proxying
for sensitive features, we need to include sensitive features in the training
phase, but exclude them in the test/evaluation phase while controlling for
their effects. We evaluate the performance of our algorithm on several
real-world datasets and show how fairness for these datasets can be improved
with a very small loss in accuracy
Near-Optimal Active Learning of Halfspaces via Query Synthesis in the Noisy Setting
In this paper, we consider the problem of actively learning a linear
classifier through query synthesis where the learner can construct artificial
queries in order to estimate the true decision boundaries. This problem has
recently gained a lot of interest in automated science and adversarial reverse
engineering for which only heuristic algorithms are known. In such
applications, queries can be constructed de novo to elicit information (e.g.,
automated science) or to evade detection with minimal cost (e.g., adversarial
reverse engineering). We develop a general framework, called dimension coupling
(DC), that 1) reduces a d-dimensional learning problem to d-1 low dimensional
sub-problems, 2) solves each sub-problem efficiently, 3) appropriately
aggregates the results and outputs a linear classifier, and 4) provides a
theoretical guarantee for all possible schemes of aggregation. The proposed
method is proved resilient to noise. We show that the DC framework avoids the
curse of dimensionality: its computational complexity scales linearly with the
dimension. Moreover, we show that the query complexity of DC is near optimal
(within a constant factor of the optimum algorithm). To further support our
theoretical analysis, we compare the performance of DC with the existing work.
We observe that DC consistently outperforms the prior arts in terms of query
complexity while often running orders of magnitude faster.Comment: Accepted by AAAI 201
- …